dependent feature
TRIP: A Nonparametric Test to Diagnose Biased Feature Importance Scores
Along with accurate prediction, understanding the contribution of each feature to the making of the prediction, i.e., the importance of the feature, is a desirable and arguably necessary component of a machine learning model. For a complex model such as a random forest, such importances are not innate -- as they are, e.g., with linear regression. Efficient methods have been created to provide such capabilities, with one of the most popular among them being permutation feature importance due to its efficiency, model-agnostic nature, and perceived intuitiveness. However, permutation feature importance has been shown to be misleading in the presence of dependent features as a result of the creation of unrealistic observations when permuting the dependent features. In this work, we develop TRIP (Test for Reliable Interpretation via Permutation), a test requiring minimal assumptions that is able to detect unreliable permutation feature importance scores that are the result of model extrapolation. To build on this, we demonstrate how the test can be complemented in order to allow its use in high dimensional settings. Through testing on simulated data and applications, our results show that the test can be used to reliably detect when permutation feature importance scores are unreliable.
- North America > United States (0.14)
- Europe > Italy (0.04)
- Asia > Middle East > Israel > Jerusalem District > Jerusalem (0.04)
Explainable Artificial Intelligence for Dependent Features: Additive Effects of Collinearity
Explainable Artificial Intelligence (XAI) emerged to reveal the internal mechanism of machine learning models and how the features affect the prediction outcome. Collinearity is one of the big issues that XAI methods face when identifying the most informative features in the model. Current XAI approaches assume the features in the models are independent and calculate the effect of each feature toward model prediction independently from the rest of the features. However, such assumption is not realistic in real life applications. We propose an Additive Effects of Collinearity (AEC) as a novel XAI method that aim to considers the collinearity issue when it models the effect of each feature in the model on the outcome. AEC is based on the idea of dividing multivariate models into several univariate models in order to examine their impact on each other and consequently on the outcome. The proposed method is implemented using simulated and real data to validate its efficiency comparing with the a state of arts XAI method. The results indicate that AEC is more robust and stable against the impact of collinearity when it explains AI models compared with the state of arts XAI method.
- Europe > United Kingdom > England > Greater London > London (0.15)
- North America > United States > California > San Francisco County > San Francisco (0.14)
- Europe > United Kingdom > England > Leicestershire > Leicester (0.05)
- (3 more...)
Cross-Domain Keyword Extraction with Keyness Patterns
Domain dependence and annotation subjectivity pose challenges for supervised keyword extraction. Based on the premises that second-order keyness patterns are existent at the community level and learnable from annotated keyword extraction datasets, this paper proposes a supervised ranking approach to keyword extraction that ranks keywords with keyness patterns consisting of independent features (such as sublanguage domain and term length) and three categories of dependent features -- heuristic features, specificity features, and representavity features. The approach uses two convolutional-neural-network based models to learn keyness patterns from keyword datasets and overcomes annotation subjectivity by training the two models with bootstrap sampling strategy. Experiments demonstrate that the approach not only achieves state-of-the-art performance on ten keyword datasets in general supervised keyword extraction with an average top-10-F-measure of 0.316 , but also robust cross-domain performance with an average top-10-F-measure of 0.346 on four datasets that are excluded in the training process. Such cross-domain robustness is attributed to the fact that community-level keyness patterns are limited in number and temperately independent of language domains, the distinction between independent features and dependent features, and the sampling training strategy that balances excess risk and lack of negative training data.
- Europe > Netherlands > North Holland > Amsterdam (0.04)
- Asia > China > Hubei Province > Wuhan (0.04)
- North America > United States > New York > New York County > New York City (0.04)
- (2 more...)
- Media (0.67)
- Transportation (0.46)
- Information Technology > Artificial Intelligence > Natural Language > Information Retrieval (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Statistical Learning (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (1.00)
Explaining the Model and Feature Dependencies by Decomposition of the Shapley Value
Michiels, Joran, De Vos, Maarten, Suykens, Johan
Shapley values have become one of the go-to methods to explain complex models to end-users. They provide a model agnostic post-hoc explanation with foundations in game theory: what is the worth of a player (in machine learning, a feature value) in the objective function (the output of the complex machine learning model). One downside is that they always require outputs of the model when some features are missing. These are usually computed by taking the expectation over the missing features. This however introduces a non-trivial choice: do we condition on the unknown features or not? In this paper we examine this question and claim that they represent two different explanations which are valid for different end-users: one that explains the model and one that explains the model combined with the feature dependencies in the data. We propose a new algorithmic approach to combine both explanations, removing the burden of choice and enhancing the explanatory power of Shapley values, and show that it achieves intuitive results on simple problems. We apply our method to two real-world datasets and discuss the explanations. Finally, we demonstrate how our method is either equivalent or superior to state-to-of-art Shapley value implementations while simultaneously allowing for increased insight into the model-data structure.
- Europe > Belgium > Flanders > Flemish Brabant > Leuven (0.05)
- Africa > Middle East > Algeria (0.04)
Balancing Explainability-Accuracy of Complex Models
Sengupta, Poushali, Zhang, Yan, Maharjan, Sabita, Eliassen, Frank
Explainability of AI models is an important topic that can have a significant impact in all domains and applications from autonomous driving to healthcare. The existing approaches to explainable AI (XAI) are mainly limited to simple machine learning algorithms, and the research regarding the explainability-accuracy tradeoff is still in its infancy especially when we are concerned about complex machine learning techniques like neural networks and deep learning (DL). In this work, we introduce a new approach for complex models based on the co-relation impact which enhances the explainability considerably while also ensuring the accuracy at a high level. We propose approaches for both scenarios of independent features and dependent features. In addition, we study the uncertainty associated with features and output. Furthermore, we provide an upper bound of the computation complexity of our proposed approach for the dependent features. The complexity bound depends on the order of logarithmic of the number of observations which provides a reliable result considering the higher dimension of dependent feature space with a smaller number of observations.
- Europe > Austria > Vienna (0.14)
- Europe > Norway > Eastern Norway > Oslo (0.05)
- North America > United States > California (0.04)
- Europe > Belgium > Flanders > East Flanders > Ghent (0.04)
- Research Report (0.51)
- Overview (0.46)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks (0.54)
- Information Technology > Artificial Intelligence > Machine Learning > Statistical Learning (0.46)
- Information Technology > Artificial Intelligence > Issues > Social & Ethical Issues (0.36)
- Information Technology > Artificial Intelligence > Natural Language > Explanation & Argumentation (0.35)
Characterizing the contribution of dependent features in XAI methods
Salih, Ahmed, Galazzo, Ilaria Boscolo, Raisi-Estabragh, Zahra, Petersen, Steffen E., Menegaz, Gloria, Radeva, Petia
Explainable Artificial Intelligence (XAI) provides tools to help understanding how the machine learning models work and reach a specific outcome. It helps to increase the interpretability of models and makes the models more trustworthy and transparent. In this context, many XAI methods were proposed being SHAP and LIME the most popular. However, the proposed methods assume that used predictors in the machine learning models are independent which in general is not necessarily true. Such assumption casts shadows on the robustness of the XAI outcomes such as the list of informative predictors. Here, we propose a simple, yet useful proxy that modifies the outcome of any XAI feature ranking method allowing to account for the dependency among the predictors. The proposed approach has the advantage of being model-agnostic as well as simple to calculate the impact of each predictor in the model in presence of collinearity.
- Europe > United Kingdom > England > Greater London > London (0.05)
- Europe > Spain > Catalonia > Barcelona Province > Barcelona (0.04)
- Asia > Middle East > Iraq > Kurdistan Region > Duhok Governorate > Zakho (0.04)
- (3 more...)
- Research Report > Experimental Study (0.47)
- Research Report > New Finding (0.46)
Causal Analysis utilizing CausalML
We often talk about correlation vs causation in theory but while implementing Data Science solutions towards solving business problems not much influence is given to validating causation amongst independent and dependent features. Conventional Machine Learning methods identify patterns in existing data to make predictions and they always result in retrieving some underlying patterns even if they are not real and factitious. The assumption is these patterns are same in training, testing, validation data and deployed environments. However, if these patterns change for some reason, ML models fail. The reasons could be numerous like a distribution shift, external unexpected factor etc. Causal Machine Learning helps by defining treatment condition with respective control data to make causal inferences guiding machine learning models to pay attention to cause and effect relations.
Multiple Linear Regression Using Python and Scikit-learn
This article was published as a part of the Data Science Blogathon. If you are on the path of learning data science, then you definitely have an understanding of what machine learning is. In today's digital world everyone knows what Machine Learning is because it was a trending digital technology across the world. Every step towards adaptation of the future world leads by this current technology, and this current technology is led by data scientists like you and me . Here we only discuss machine learning, If you don't know what it is, then we take a brief introduction to it: Machine learning is the study of the algorithms of computers, that improve automatically through experience and by the use of data. This is the simple definition of machine learning, and when we go into deep then we find that there are huge numbers of algorithms that are used in model building.
Pitfalls to Avoid when Interpreting Machine Learning Models
Molnar, Christoph, König, Gunnar, Herbinger, Julia, Freiesleben, Timo, Dandl, Susanne, Scholbeck, Christian A., Casalicchio, Giuseppe, Grosse-Wentrup, Moritz, Bischl, Bernd
Modern requirements for machine learning (ML) models include both high predictive performance and model interpretability. A growing number of techniques provide model interpretations, but can lead to wrong conclusions if applied incorrectly. We illustrate pitfalls of ML model interpretation such as bad model generalization, dependent features, feature interactions or unjustified causal interpretations. Our paper addresses ML practitioners by raising awareness of pitfalls and pointing out solutions for correct model interpretation, as well as ML researchers by discussing open issues for further research.
- Europe > Austria > Vienna (0.14)
- Europe > Germany > Bavaria > Upper Bavaria > Munich (0.05)
- Asia > Middle East > Jordan (0.04)
- (2 more...)
Personalized Influence Estimation Technique
Pathak, Kumarjit, Kapila, Jitin, Barvey, Aasheesh
Customer Satisfaction is the most important factors in the industry irrespective of domain. Key Driver Analysis is a common practice in data science to help the business to evaluate the same. Understanding key features, which influence the outcome or dependent feature, is highly important in statistical model building. This helps to eliminate not so important factors from the model to minimize noise coming from the features, which does not contribute significantly enough to explain the behavior of the dependent feature, which we want to predict. Personalized Influence Estimation is a technique introduced in this paper, which can estimate key factor influence for individual observations, which contribute most for each observations behavior pattern based on the dependent class or estimate. Observations can come from multiple business problem i.e. customers related to satisfaction study, customer related to Fraud Detection, network devices for Fault detection etc. It is highly important to understand the cause of issue at each observation level to take appropriate Individualized action at customer level or device level etc. This technique is based on joint behavior of the feature dimension for the specific observation, and relative importance of the feature to estimate impact. The technique mentioned in this paper is aimed to help organizations to understand each respondents or observations individual key contributing factor of Influence. Result of the experiment is really encouraging and able to justify key reasons for churn for majority of the sample appropriately